Goto

Collaborating Authors

 shadow mask



Supplementary Material for " AllClear: A Comprehensive Dataset and Benchmark for Cloud Removal in Satellite Imagery "

Neural Information Processing Systems

In Sec. 2 we include a We include a datasheet for our dataset following the methodology from "Datasheets for Datasets" Ge-17 In this section, we include the prompts from Gebru et al. [2021] in blue, and in For what purpose was the dataset created? Was there a specific task in mind? The dataset was created to facilitate research development on cloud removal in satellite imagery. Specifically, our task is more temporally aligned than previous benchmarks. Who created the dataset (e.g., which team, research group) and on behalf of which entity (e.g., Who funded the creation of the dataset?



Supplementary Material for " AllClear: A Comprehensive Dataset and Benchmark for Cloud Removal in Satellite Imagery "

Neural Information Processing Systems

In Sec. 2 we include a We include a datasheet for our dataset following the methodology from "Datasheets for Datasets" Ge-17 In this section, we include the prompts from Gebru et al. [2021] in blue, and in For what purpose was the dataset created? Was there a specific task in mind? The dataset was created to facilitate research development on cloud removal in satellite imagery. Specifically, our task is more temporally aligned than previous benchmarks. Who created the dataset (e.g., which team, research group) and on behalf of which entity (e.g., Who funded the creation of the dataset?


AllClear: A Comprehensive Dataset and Benchmark for Cloud Removal in Satellite Imagery

Zhou, Hangyu, Kao, Chia-Hsiang, Phoo, Cheng Perng, Mall, Utkarsh, Hariharan, Bharath, Bala, Kavita

arXiv.org Artificial Intelligence

Clouds in satellite imagery pose a significant challenge for downstream applications. A major challenge in current cloud removal research is the absence of a comprehensive benchmark and a sufficiently large and diverse training dataset. To address this problem, we introduce the largest public dataset -- $\textit{AllClear}$ for cloud removal, featuring 23,742 globally distributed regions of interest (ROIs) with diverse land-use patterns, comprising 4 million images in total. Each ROI includes complete temporal captures from the year 2022, with (1) multi-spectral optical imagery from Sentinel-2 and Landsat 8/9, (2) synthetic aperture radar (SAR) imagery from Sentinel-1, and (3) auxiliary remote sensing products such as cloud masks and land cover maps. We validate the effectiveness of our dataset by benchmarking performance, demonstrating the scaling law -- the PSNR rises from $28.47$ to $33.87$ with $30\times$ more data, and conducting ablation studies on the temporal length and the importance of individual modalities. This dataset aims to provide comprehensive coverage of the Earth's surface and promote better cloud removal results.


Latent Feature-Guided Diffusion Models for Shadow Removal

Mei, Kangfu, Figueroa, Luis, Lin, Zhe, Ding, Zhihong, Cohen, Scott, Patel, Vishal M.

arXiv.org Artificial Intelligence

Motivated by the success of diffusionbased Recovering textures under shadows has remained a challenging image restoration models [38, 41], we adapt diffusion problem due to the difficulty of inferring shadowfree models for the task of shadow removal by conditioning on scenes from shadow images. In this paper, we propose the input shadow image and corresponding shadow mask as the use of diffusion models as they offer a promising approach a baseline approach to generate shadow-free images. However, to gradually refine the details of shadow regions preserving and generating high-fidelity textures and during the diffusion process. Our method improves this colors in the shadow region after removal is non-trivial. The process by conditioning on a learned latent feature space baseline model appears to favor borrowing textures from that inherits the characteristics of shadow-free images, thus the surrounding non-shadow areas rather than focusing on avoiding the limitation of conventional methods that condition restoring the original details underneath the shadow, which on degraded images only. Additionally, we propose results in incorrect color mixtures and loss of detail in the to alleviate potential local optima during training by fusing shadow region. In Figure 1, we show one of the representative noise features with the diffusion network. We demonstrate issues of image-mask conditioning, i.e., the model synthesizes the effectiveness of our approach which outperforms results containing an incorrect color mixture.


Decomposer: Semi-supervised Learning of Image Restoration and Image Decomposition

Meinardus, Boris, Trzeciakiewicz, Mariusz, Herzig, Tim, Kwiatkowski, Monika, Matern, Simon, Hellwich, Olaf

arXiv.org Artificial Intelligence

We present Decomposer, a semi-supervised reconstruction model that decomposes distorted image sequences into their fundamental building blocks - the original image and the applied augmentations, i.e., shadow, light, and occlusions. To solve this problem, we use the SIDAR dataset that provides a large number of distorted image sequences: each sequence contains images with shadows, lighting, and occlusions applied to an undistorted version. Each distortion changes the original signal in different ways, e.g., additive or multiplicative noise. We propose a transformer-based model to explicitly learn this decomposition. The sequential model uses 3D Swin-Transformers for spatio-temporal encoding and 3D U-Nets as prediction heads for individual parts of the decomposition. We demonstrate that by separately pre-training our model on weakly supervised pseudo labels, we can steer our model to optimize for our ambiguous problem definition and learn to differentiate between the different image distortions.


Towards Learning Neural Representations from Shadows

Tiwary, Kushagra, Klinghoffer, Tzofi, Raskar, Ramesh

arXiv.org Artificial Intelligence

We present a method that learns neural shadow fields which are neural scene representations that are only learnt from the shadows present in the scene. While traditional shape-from-shadow (SfS) algorithms reconstruct geometry from shadows, they assume a fixed scanning setup and fail to generalize to complex scenes. Neural rendering algorithms, on the other hand, rely on photometric consistency between RGB images, but largely ignore physical cues such as shadows, which have been shown to provide valuable information about the scene. We observe that shadows are a powerful cue that can constrain neural scene representations to learn SfS, and even outperform NeRF to reconstruct otherwise hidden geometry. We propose a graphics-inspired differentiable approach to render accurate shadows with volumetric rendering, predicting a shadow map that can be compared to the ground truth shadow. Even with just binary shadow maps, we show that neural rendering can localize the object and estimate coarse geometry. Our approach reveals that sparse cues in images can be used to estimate geometry using differentiable volumetric rendering. Moreover, our framework is highly generalizable and can work alongside existing 3D reconstruction techniques that otherwise only use photometric consistency.


Self-Supervised Shadow Removal

Vasluianu, Florin-Alexandru, Romero, Andres, Van Gool, Luc, Timofte, Radu

arXiv.org Artificial Intelligence

Shadow removal is an important computer vision task aiming at the detection and successful removal of the shadow produced by an occluded light source and a photo-realistic restoration of the image contents. Decades of re-search produced a multitude of hand-crafted restoration techniques and, more recently, learned solutions from shad-owed and shadow-free training image pairs. In this work,we propose an unsupervised single image shadow removal solution via self-supervised learning by using a conditioned mask. In contrast to existing literature, we do not require paired shadowed and shadow-free images, instead we rely on self-supervision and jointly learn deep models to remove and add shadows to images. We validate our approach on the recently introduced ISTD and USR datasets. We largely improve quantitatively and qualitatively over the compared methods and set a new state-of-the-art performance in single image shadow removal.